| Patrice Riemens on Fri, 13 Mar 2009 13:14:33 -0400 (EDT) |
[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]
| <nettime> Ippolita Collective: The Dark Side of Google (Chapter 2, first part) |
Dear Nettimers,
There will be a short interruption in the translation flow as I am taking
the train to Chennai/Madras and Madurai this afternoon, back on tuesday
morning. So long!
Cheerio,
patrizio & Diiiinooos!
(going to visit http://www.antennaindia.org)
-----------------------------------------------------------------------------
NB this book and translation are published under Creative Commons license
2.0
(Attribution, Non Commercial, Share Alike).
Commercial distribution requires the authorisation of the copyright
holders:
Ippolita Collective and Feltrinelli Editore, Milano (.it)
Ippolita Collective
The Dark Side of Google (continued)
Chapter 2 BeGoogle!
Google's Brain-drain or the War for Control of the Web
"I'll #%$&*^! this Google &^%$#@! Eric Schmidt is a $%&*^#! and I'll bury
him alive, like I did with other %#*(@&^! like him!". Thus foulmouthed
Microsoft's CEO Steve Balmer when he learned in May 2005 that Google had
just headhunted Kai-Fu Lee, a high ranking employee of his, and key-man of
'Redmond' for China. Kai-Fu was the one who had developed MSN Search
(engine) for the 100 million Chinese Microsoft users. Balmer's expletives
were of course targeted at his opposite number at Google, a former Sun
Microsystems and Novell honcho, firms also that Microsystem had battled
with before, both on the market and in court. Kai-Fu Lee was boss of the
MS research lab near Shanghai.
Microsoft immediately started a court case against its former employee and
Google specifically, accusing Kai-Fu Lee of violating extremely
confidential contractual agreements existing between the Redmond and
Mountain View rivals. Microsoft lawyers argued that Kai-Fu Lee, as
executive director, must be in the know of MS industrial and trade
secrets, and would not hesitate to put these technologies and the social
network and economic know-how he had accrued at MS to use to bolster the
profits of the competitor's firm. This contentious personage didn't come
cheaply by the way. His entry 'salary' amounted to US$ 2,5 million, with
20.000 Google shares as a side perq. Exorbitant figures which give some
idea of the wager at stake - and we were not only talking about the
Chinese market.
The lawsuit between the two giants was finally settled out of court in
December 2005 - with just one month left before the case was to come up.
The particulars of the deal are completely confidential. May be large
amounts of money have changed hands, or may be Microsoft managed to force
Kai-Fu Lee to keep mum about anything he knew from his previous
employment.
This story is merely one of the most curious and emblematic illustrations
of a trend that had become noticeable for a few years now: Kai-Fu Lee was
actually only the umpteenth senior employee that had switched to Google,
"the firm that looks more and more like Microsoft" as Bill Gates had
loudly complained. Bill himself was left in a cleft shtick as he faced the
nasty choice of either diabolising the two student prodigies - reinforcing
thereby their image as his 'kind and generous' opponents in the world of
IT - or to pretend that they did not really matter and were not very much
worth of attention as competitors .
The truth is that Bill Gates knew all too well how much a switch-over of
managers means for a firm's core business, especially in the IT sector:
Microsoft had often enough made use of the same trick against its own
competitors. The commercial tactic consisting in headhunting key personnel
of rival firms in order to tap their industrial secrets and their
production and resources management know-how has always been part and
parcel of industrial competition. But in the era of the information
economy, the practice has become markedly more prevalent, and more
diffuse.
So this management choice of Brin's and Page's clearly indicates what
Google's ultimate aims are: to become the Web's most comprehensive and
customizable platform, by adapting all its services to the singular needs
of each of its users, and this together with maintaining an immense
reservoir of information. To put it simply, Google is pushing full speed
ahead to catalogue every type of digital information, ranging from
websites to discussion groups, picture galleries, e-mails, blogs, and
whatever you can think of, without any limit in sight. This amounts to
open war with Microsoft, whose Internet 'Explorer' browser, MSN portal,
and its Hotmail e-mail service, makes it after all, and for the time
being, Google's principal foe [competitor].
The overlap between the domains of interest of both firms is growing by
the day: both aspire to be the one and only medium to access whichever
digital activity. Microsoft has achieved predominance by imposing its
Windows operating system, its Office software suite {and its Explorer
browser} as the current computing standard both at work and at home. On
its side, Google has been profiling itself as the global number one
mediator of web services, especially with regard to search, its core
business, offered in all possible formats, but also with particular
ancillary services such as e-mail ('GMail'). At the risk of
simplification, one could say that Microsoft has been for years in a
dominant position thanks to products that pertain to services, whereas
Google is now seeking dominance through services running on products.
The outcome of this competition, hence, is dependant on users' choices and
on the future standards Google wants to impose. Developing certain web
programmes intended to funnel requests for services only through the
browser amounts to deny a market to those who have always invested heavily
in products and in creating new operating software architecture. [French
text unclear here, I guess the gist is: Google's going to literally
vaporize all 'static' M$ products by going full tilt for the 'Internet in
the clouds' paradigm, cf. next sentence -TR]. The same holds true for
/markets in/ the economy at large: there is a shift from a wholesaler/mass
market approach (Microsoft), trying to sell licenses of one and the same
product or service, to a completely customised one, where products can be
downloaded from the web.
Long tails in the Net. Google vs. Microsoft in the economy of search
Google's second line of argument is based on the key point John Batelle
made in his numerous writings [*N2]: the ascent of the 'economy of
search'. In his essay "The Second Search", Batelle, who is a journalist
and counted amongst the founders of WIRED magazine, argues that the future
of on-line commerce lies with personalised searches paid by the users
[themselves?]. Google, which is sitting on top of the largest data-bank of
'search intentions' by users, finds itself in the most advantageous
position to make this possible, thanks to its very finely ramified
network, made up on one side by a famously efficient advertising platform
(AdWords) and on the other, by a bank of advertisers (AdSense) now good
for several millions of websites. Google's wager is that it will be able
to satisfy any wish the users may express through their search query, by
providing new services geared towards 'consumerism at the individual
level'. Each and every user/customer will hit exactly what she/he wants,
the product that is precisely geared to her/his needs. The best known of
these 'mass personalised' online services is the one offered by
Amazon.com, which is well on its way to make far more money out of selling
books or Cd's one at a time to individual customers than to pile up
hundreds or even thousand of copies of a best seller. The numerous
customers buying not particularly well-selling books online constitute a
myriad of 'events' infrequently occurring in themselves, and sometimes
even only once. To be able to satisfy nevertheless such 'personalised
searches' is the secret of Amazon.com's distribution power. It would be
impossible for a traditional book-seller, whose operational model is
stacked on shops, stocks, and limited orders, to have the ease of delivery
of million of titles at once Amazon.com has: most of its revenues has to
come from novelties and best-sellers. Selling one copy of a book to a
single customer is not profitable for a traditional bookshop, but is is
for Amazon.com, which capitalizes on the 'economy of search' of the
'online marketplace'.
This type of market is called 'long tail' in {new} economic parlance
[*N3]. The theory of 'long tails' goes at least back to 'Pareto's
distribution' [*N4], where "there a few events that have a high
occurrence, whereas many have a low one". Statistically such distribution
is represented by a hyperbole {graph} where the 'long tail' is made up of
a myriad of events that are pretty much insignificant in themselves, but
which taken together, represent a considerable sum. Mathematically
speaking, a 'long tail' distribution follows the pattern of what is called
"power's law" [*N5].
The 'winning strategy' in a long tail market hence is not to lower prices
on the most popular products, but to have a wider range of offerings. This
makes it possible to sell 'searchlight products' while selling few items
at a time, but out of a very large range of different products.
Commercially speaking, it turns out that the highest sales occur in the
realm of small transactions. The largest part of sales on the net is a
long tail phenomenon. Google makes turnover by selling cheap
advertisements to millions of users with text ads, not by selling a lot of
advertising space in one go to a few big firms for a hefty fee.
Batelle takes interest in the application of search into not yet explored
markets. In the case of Google, the enormous amount of data that is
available in order to make searches possible is what has made the milking
of the 'long tail' possible. In the domain of e-commerce, long tails have
three consequences: first, thanks to the Internet, it becomes possible for
non-so-frequently asked products to collectively represent a larger market
than the one commanded by the small number of articles that do enjoy large
sales; second, the Internet favors the proliferation of sellers - and of
markets (such as is illustrated by the auction site eBay); and thirdly,
thanks to search, the shift from {traditional,} mass market to that of
niches becomes a realistic scenario.
This last tendency finds its origin in the spontaneous emergence of groups
{of like-minded people}, something occurring on a large scale in networks.
On the Internet, even the most important groups by number are not
necessarily made up of homogeneous masses of individual people, but rather
of colourful communities of users banding together because of a shared
passion, or a common interest or goal. The opposition between niche and
mass therefore is not very relevant to the identification of the segment
of the market that is aimed at. From a commercial point of view, this
leads to the creation of e-commerce sites for products attractive only to
a very specific type of potential customers, who would never have
constituted a profitable market outside online distribution. Take for
instance typically geeky tee-shirts, or watches giving 'binary' time,
flashy computer boxes or other must-have items targeted at the techie
crowd. The amplitude of the supply makes good for the narrowness of the
demand, which is spread over a very extensive range of highly personalised
products. An interesting article by Charles H. Ferguson [*N6] points out
that in such a scenario, it is most likely that Google and Microsoft will
confront each other for real for the control [monopoly?] of indexing,
searching, and data-mining, and this over the full spectrum of digital
services and devices. Microsoft is now massively investing in web
services: in November 2004 it launched a beta version of a search engine
that would answer queries made in everyday language, and return answers
that would be personalised according to geographical location of the user;
in February 2005, this MSN Search engine was improved further [*N7]. with
MSN Search, it becomes possible to check out Encarta, Microsoft's
multimedia cyclopedia. But for the time being, browsing is limited to two
hours, with a little watch window telling you how much time remains ...
Microsoft has thus decided to develop its own web search system on PCs,
without resorting to Google, despite the fact that the latter has been for
years now #1 in the search business (with Yahoo! as sole serious
competitor).
Taken as a whole, it would appear that the markets that are linked to the
economy of search are much larger than the existing markets for search
services as such. Microsoft is undoubtedly lagging behind in this area,
but the firm from Redmond might wel unleash {its trademark} savage
strategies, which would be difficult for Google to counter. It could for
instance take a loss on investments, integrate its search engine to its
Explorer browser and offer the package for free, or start a price war on
advertisements and so starve its competitor of liquidity. And in the
meanwhile, the new Windows Vista operating system developed in Redmond is
supposed to offer innovative search options [looks like fat chance...;-)
-TR]. Also take note that Microsoft was lagging very much behind Netscape
(the first web browser that was freely down-loadable) in its time , and
yet Explorer managed to displace {and dispatch} it - and not really
because it was so much better! But if Microsoft indeed has a long
experience of the market and also very deep pockets, Google has not a bad
hand either. It is the very incarnation of the young, emergent enterprise,
it has built up a reputation as a firm that is committed to research and
technical excellence, it preaches the gospel of speed with regard to
users' search satisfaction and does so with nifty and sober interfaces, in
one word it imposes itself by simply being technically the best search
engine around. In the battle for control of the Web Google appears to have
a slight advantage. However, one should not forget that Microsoft's range
of activity is without par since it covers not only the Web but the whole
gamut of information technologies, from tools like the Windows operating
system or the MS Office suite, to contents like Encarta, and hi-end
research platforms like dotNet, etc. Given the wager at stake - basically
the access to any kind of digital piece of information, and the profits
deriving from it - peaceful cohabitation {between the two giants} looks
unlikely. Google is still in the race for now - but for how long? [MR/FCG:
" but (G) won't be able to stand up very long"]
The War of Standards
Let's follow up on Ferguson's argument: the story starting now is a war of
standards. Three actors are in the game, for now: Google, Yahoo!, and
Microsoft. The industry of search, as Batelle pointed out also, is growing
at a fast pace. Emerging technologies, or the ones that are currently
under consolidation - think broadband enabled audio & video streaming for
instance, or VoIP telephony (e.g. Skype, or Google's GTalk), or instant
messaging - all are generating Himalayas of data still waiting for proper
indexation. And proper 'usabilitization' for the full spectrum of new
electronic vectors like pals, gsm's, audio-video devices, satellite
navigators, etc - all these being interlinked for the satisfaction of
users - but all milked in the end as supports for intrusive advertising.
In order for these tools to be compatible with all kinds of different
systems and which each other, new standards will be necessary, and their
introduction /in the market/ is unlikely to be a painless process. What
causes a war of standards is the fact that technology markets demand
common languages in order to organise an ever-increasing complexity. The
value of information lies in its distribution; but it is easier to spread
around real tokens [? analog stuff?] than audio, or worse still, video
documents: the heavier the data-load, the more powerful 'pipes' it
requires and the more demands in puts on the way in which the information
is managed [French text somewhat unclear here]. Traditionally, legal
ownership of a crucial technology standard has always been the source of
very comfortable revenues [*N8]. It has indeed happened that the adoption
of an open, non proprietary standard - such as the 'http' protocol -
created a situation that is beneficial to all parties. But often, the
dominant solutions are not qualitatively the best, as "it is often more
important to be able to rely on a winning marketing strategy". This being
said, there are a number of trends emerging regarding the winners. They
usually sell platforms that work everywhere irrespective of the hardware,
like for instance Microsoft's operating systems, and this as opposed to
the closely integrated hardware and software solutions offered bay Apple
or Sun Microsystems. Winning architectures are proprietary and difficult
to reduplicate, yet they are at the same time very 'open', that is, they
propose publicly accessible interfaces so that can be developed further by
independent programmers, and in the end by the users themselves. In doing
so, the architecture in question is able both to penetrate all markets and
at the same time create a situation of attraction and 'lock-in'. Or to put
it differently, it pulls the users towards a specific architecture, and
once in, it becomes next to impossible to switch to a competing system
without incurring great difficulties and huge expenses [*N9]. The aim:
impose a closed standard, and obtain a monopoly.
A very clear illustration of the battle for hegemony in the domain of
standards is the challenge Skype and GTalk are throwing at each other. For
the time being, Skype enjoys a position of near-monopoly on VoIP for
domestic use. Yet it is possible that it has not well estimated the time
needed by development communities to assimilate [embrace?] open
technologies. Till not so long ago, Skype was the only solution that
really worked for anyone wanting to put phone calls through the Internet -
even if {the person was} technically clueless. Skype's proprietary
technologies, however, could well be overtaken by GTalk, which is entirely
based on F/OSS (and especially on the 'Jabber' communication protocol),
all offered with development libraries under copyleft licenses, something
that attracts a lot of creative energies in the project as coders vie to
increase the punch of Google's VoIP network. In which case the adoption of
F/OSS would prove to be the winning strategy and erode Skype's dominance.
Of course, Skype could then choose to make its own codes public in order
to tilt the ante back in its favor. The choice between adoption of
proprietary technologies and platforms, closing access - but keeping
public the development interfaces, and going for openness, is therefore of
paramount importance in the strategy of control of the Web, and of the
economy of search in general.
But access to the economy of search is already closed, or 'locked-in', as
economists would say: that any new entrant, or 'start-up', could ever
compete with Yahoo! or Google in the indexation of billions of web-pages
is clearly unthinkable. Even if such a firm had a better algorithm for its
spider, the investments needed would be prohibitive. Yet there are a lot
of side-aspects, especially with regard to the interface between various
search systems, which lend themselves for a bevy of 'mission critical',
yet affordable innovations. This is for instance the case with
'libraries', small pieces of software which make it possible to link up
heterogeneous systems together and function as 'translators' between
systems, languages, and search engines. Together with integration
methodologies between various arrangements and ways to share data and
search results, they represent areas that could be developed by
individual, independent researchers rather than by large companies [*N10].
We will later into more details into the issue of interfaces and
libraries.
For now, it is important to note that none of the players in this game is
in a position of absolute dominance, something we can be thankful for.
Imagine what the situation would be if a complete monopoly of search, by
whatever private actor, existed by virtue of its factual imposition of one
standard. Obviously, the first problem to arise would be the issue of
privacy: who would own the indexed data on which searches would take
place, reaping humongous profits in the process ? Moreover, since it is
already possible to tap into quite an unbelievable amount of information
just by typing the name of an individual in the Google search window, and
since in a near future the quality and quantity of such information will
not only be greatly increased, but even further augmented by the
possibility to cross-search among heterogeneous data, one can assume that
the control exercised on individuals will become ever more suffocating and
totalitarian: it will cross-aggregate confidential data with medical
records, phone conversations, e-mails, pictures, videos, blogs and opinion
pieces, and even ADN info. Google would then become the premiere access
point to the digital panopticon [*N11].
So let's us now have a look at the weapons that are deployed in this very
real war for the control of the networks.
(to be continued)
--------------------------
Translated by Patrice Riemens
This translation project is supported and facilitated by:
The Center for Internet and Society, Bangalore
(http://cis-india.org)
The Tactical Technology Collective, Bangalore Office
(http://www.tacticaltech.org)
Visthar, Dodda Gubbi post, Kothanyur-Bangalore
(http://www.visthar.org)
# distributed via <nettime>: no commercial use without permission
# <nettime> is a moderated mailing list for net criticism,
# collaborative text filtering and cultural politics of the nets
# more info: http://mail.kein.org/mailman/listinfo/nettime-l
# archive: http://www.nettime.org contact: nettime@kein.org